Logo

Logo

Top AI experts, CEOs issue 22-word warning against ‘risk of extinction’ from AI

Top researchers, experts and CEOs (including Sam Altman of OpenAI) have issued a fresh warning about the existential threat the artificial intelligence (AI) poses to humanity.

Top AI experts, CEOs issue 22-word warning against ‘risk of extinction’ from AI

Representational Image (Photo: Pixabay.com)

Top researchers, experts and CEOs (including Sam Altman of OpenAI) have issued a fresh warning about the existential threat the artificial intelligence (AI) poses to humanity.

In a 22-word statement, they said that “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”

The statement was published by US-based nonprofit, the Center for AI Safety, and was co-signed by Google DeepMind CEO Demis Hassabis as well as Geoffrey Hinton and Youshua Bengio, two of the three AI researchers who won the 2018 Turing Award for their work on AI.

Advertisement

“AI experts, journalists, policymakers, and the public are increasingly discussing a broad spectrum of important and urgent risks from AI. Even so, it can be difficult to voice concerns about some of advanced AI’s most severe risks,” the Center said.

The statement “aims to overcome this obstacle and open up discussion”.

It is also meant to create common knowledge of the growing number of experts and public figures who also take some of advanced AI’s most severe risks seriously, the stakeholders said.

In March this year, several top entrepreneurs and AI researchers, including Tesla CEO Musk and Steve Wozniak, Co-founder of Apple, wrote an open letter, asking all AI labs to immediately pause training of AI systems more powerful than GPT-4 for at least six months.

Last week, Altman said that now is a good time to start thinking about the governance of superintelligence — future AI systems dramatically more capable than even artificial general intelligence (AGI).

Altman stressed that the world must mitigate the risks of today’s AI technology too, “but superintelligence will require special treatment and coordination”.

“Given the possibility of existential risk, we can’t just be reactive. Nuclear energy is a commonly used historical example of a technology with this property; synthetic biology is another example,” he noted.

Altman earlier admitted that if AI technology goes wrong, it can go quite wrong, as US senators expressed their fears about AI chatbots like ChatGPT.

Advertisement